Development of Technical
ASSISTANCE MANUAL for Energy Simulation Tool Approval
 
Draft Meeting Minutes
October 17, 2011
1:00 P.M. – 3:00 PM.
Department of Community Affairs
Building Codes and Standards 
2555 Shumard Oak Boulevard 
Tallahassee, Florida 32399-2100 
(850) 487-1824
Meeting Purpose:
As
part of the new 2010 Florida Energy Code, the Commission will be charged with
the responsibility of approving energy simulation tools (tool). 
Under
the proposed service, the Contractor, JM Jadu Corp, will develop the procedure
(aka Manual) for reviewing and validating the tools.  
Meeting Objectives:
Review
the first draft of the technical assistance manual.
Meeting commences with roll call and confirmation of quorum of the workgroup members.
The agenda for the current meeting and minutes of first meeting was approved without any changes. Mr. Madani explained that the name of the manual is now called the “Technical Assistance Manual” and is available for vendor to use to submit their software tool.
Using a comments matrix emailed and available on the website, www.floridabuilding.org, Mr. Jadunandan reviewed the comments which came in from primarily from two groups. The first comments concerned the term “equivalent results” to “similar results.” It was concluded that compliance software either pass or fail the tests. It was agreed to stricken the entire sentence. The comments regarding time value dependent testing results were also stricken from the manual. It was agreed that submitting content about failed test results is not applicable. It was concluded that compliance software either pass or fail the tests.
The next comment is with respect to the type of software testing used in this manual. The testing is based on inter-model comparisons and is one portion of an overall validation methodology that was first developed by NREL in 1983 (Judkoff et al 1983). The draft listed the three types of evaluation in the overall methodology. This manual uses the comparative testing method only. It was agreed that only this information is necessary in the manual. The comparative testing is when a program is compared to itself or to other programs.
Mr. Jadunandan informed the workgroup that the detailed testing items 5 thru 9 will be removed. On test item number 3, the only data available for Florida is the Florida Hers Best Test results. Only the original weather file and complete coherent package for Orlando is available and is the choice of the workgroup. Las Vegas and Colorado Springs was suggested as a possible additional comparison results. Las Vegas may be used for heating and Colorado Springs cooling. Reference results for Las Vegas and Colorado Springs are not provided by the State.
There was some concern that only one data set is available. Mr. Jadunandan informed the group currently the State is will not conduct any new tests to generate new reference results. Mr. Neymark informed the group that most tools have ran the Orlando, Las Vegas, and Colorado Springs to confirm to Resnet requirements.
It was agreed that all three locations will be required by all vendors.
The other comment was regarding tier1 and tier2 testing. It was agreed that only tier1 is necessary. Tier2 is primarily passive solar testing and building orientations. It was agreed that tier2 testing be stricken from the manual.
There is a new set of reference of ASHRAE 140-2011 may be available. The group agreed to use the ASHRAE 140-2007 as the reference material instead of ASHRAE 140-2004.
Mr. Neymark explains the reasons to use logical explanation if a vendor results falls just outside of the acceptable range. The discussion continues about COMNET references and auto generation testing. Mr. Swami express concern about auto-generation and the Mr. Madani states that the Vendors and State will continue to work on testing cases and results for the future revision of the manual. Mr. Swami explains that testing is available for performance testing, but not for the generating the code-compliant reference design. The State agrees that currently there is no auto-generation test, but will continue to encourage the vendors to develop these test.
It was suggested that the State provide a design test case and have the Vendor provide an auto-generated reference design. The State can then manually checked the results to ensure that the reference design is generated correctly. All agreed that is it an important area for the State to fund and continue working on.
It was agreed that the Vendor shall include a statement that the software is capable of auto-generating the reference design using only the user input of the proposed design. This requirement is stated in the building code.
The next comment was regarding the change in the formula in Appendix A from N-1 to N. Mr. Neymark explained the reasoning for changing the formula and how it was modified to reflect the same results. Further discussions about his included the possibility that the reference results will longer be applicable. However, since the Vendors are no longer generating reference results for other cities, the entire reference is not necessary. It was agreed to remove appendix A for the draft since the vendor is only using data previously generated.
Mr. Jadunandan state that appendix B is informational and all agreed that it should be reference only. Appendix B will be deleted. Both appendixes will be replaces with templates of Forms 405, 506 and EPL Card.
The comment regarding Forms (405, 506 and EPL) referenced in the Florida Building Code, Energy Gauge. It was agreed to use FSEC generated forms as templates.
The comment was made concerning the software vendor releasing their algorithm. No define conclusion was reached on if the Vendor should release their algorithm.
Concerns were raised about how the State will handle code changes and compliance tool approval. It was noted that the manual already addresses these types of changes.
The question about Public Notification of issues and actions was asked. Mr. Madani said that this done by case by case situation and thru public proceedings via the Commissions.
Question about who is doing the review of the application. Mr. Madani answered that staff and general public have access to the vendor certification. The process is self-certification. Mr. Philip suggested it’s more of a vendor-testing. Mr. Madani state that staff will do spot check of the application. Vendor certification is open to peer review.
Mr. Philip questions why the software is responsible of sizing the HVAC (Section 5.1 2nd bullets). It was agreed to stricken this bullet item. It is required by code, but not needed as a requirement in the compliance software. This process is done by the HVAC contractor and plan reviewers. The only potential issue is some “gaming the system” and some plan reviewer may not catch it. May have text that suggests that unmet hours must be met. Some suggestions were 0 or 1%. No definite replacement text/number, if any, was agreed upon.
Concern was raised about some approved software generating slightly different results on a similar design. The users may end up purchasing all approved compliance software tool and use the one which generate the desire results. The idea of having algorithms available may provide a solution and question of who is approving the application and software. Mr. Madani suggested that the State will deal with the issues as they arise.
Florida Department of Business and Professional Regulation
FLORIDA BUILDING CODE AND
STANDARDS
 
Review 1st Draft of the
Technical Assistance Manual for Energy Simulation Tool Approval 
Post Workgroup meeting of October 17, 2011
AM: Approved as amended
| 
   Page #  | 
  
   Section  | 
  
   Comments  | 
  
   Proposed Changes   | 
  
   Remarks  | 
 ||
| 
   6  | 
  
   2  | 
  
   Joel Neymark (JN)  comment: The approval tests minimize differences in interpretation by
  providing explicit detailed descriptions of the test buildings that must be
  analyzed. For differences in the compliance software's algorithms, the
  Commission allows algorithms that yield equivalent results to those of
  the reference programs that are identified for each designated test suite.  | 
  
   The
  approval tests minimize differences in interpretation by providing explicit
  detailed descriptions of the test buildings that must be analyzed.   | 
  
   See underline text Oct. 17, 2011 WG: AM - remove text marked in
  red.  | 
 ||
| 
   19  | 
  
   5.1  | 
  
   JN comment: For tests that DO NOT COMPLY, the vendor shall supply
  diagnostic output that indicates noncompliance and gives the TDV [define
  TDV] energy information needed to evaluate the test criteria  | 
  
   
  | 
  
   See underline text Oct. 17, 2011 WG: AM – delete paragraph as noted
  in red.  | 
 ||
| 
   21  | 
  
   6.1  | 
  
   JN comment: the bulletined
  material is updated in Annex B23 of 140-2007 Addendum C or 140-2011 will have
  same – ASHRAE should publish these later this month, but we can figure out a
  way to get you internal version if you need them sooner. – or otherwise use
  the updated bullets from Judkoff and Neymark 2006] · Analytical Verification – in which the output from a program, subroutine, or algorithm is compared to the results from a known analytical solution for isolated heat transfer mechanism under simple boundary conditions · Empirical Validation – in which calculation results from a program, subroutine, or algorithm, is compared to monitored data from a real structure, test cell, or laboratory experiment · Comparative testing – in which a program is compared to itself or to other programs. The comparative approach included “sensitivity testing” and “intermodal comparisons.”  | 
  
   Replace bulletined text with text
  depicted below: Empirical Validation—in which calculated results from a program,
  subroutine, algorithm, or software object are compared to monitored data from
  a real building, test cell, or laboratory experiment.  • Analytical
  Verification—in which outputs from a program, subroutine, algorithm, or
  software object are compared to results from a known analytical solution or a
  generally accepted numerical method for isolated heat transfer under very
  simple, highly constrained boundary conditions.  • Comparative Testing—in which a
  program is compared to itself or to other programs.   | 
  
   updated bullets from Judkoff and Neymark 2006 Oct. 17, 2011 WG:  AM – restructure section so that validation
  analysis is limited to “comparative testing”.  | 
 ||
| 
   23  | 
  
   | 
  
   JN comment: edits per Barnaby, Fairey,
  Judkoff, Neymark emails of 10/4 and 10/5. Anything other than running Florida
  HERS BESTEST for only the Orlando climate is not workable, unless the Florida
  Commission wants to fund a research project to generate new reference results
  for other climates], Bill Wright:  The intent of this section is reasonable –
  to show that the proposed compliance software complies with some accepted
  standard. Items 1) and 2) on page 22 make use of work described in the
  reference document (from 1997 work by NREL) on page 22. Unfortunately, the
  Tier 1 tests are the only ones we can find that any software vendors have
  compared to. Tier 2 tests are unnecessarily extreme, apply to a very small
  group in the market, and require a scope of effort out of proportion to the
  existing or likely future market. Tier 1 tests in Orlando will achieve the
  intent of this procedure.  Item 3) asks vendors to use Appendix A in
  the Manual to “calculate the acceptable range”. This is an unnecessary
  complexity to burden vendors when the State should provide the acceptable
  range. Unfortunately, at least one of the three programs are no longer available and two of the three are longer supported. In addition, TMY3 data could be used for the 9 cities other than Orlando, but original data are not available. In addition, asking vendors to compute acceptable ranges for 9 cities multiplied by all the cases required for Tier 1 has the same problems discussed in the preceding section discussing item 3). Any criteria for acceptance should be supplied by the State, not generated by vendors. Items
  5) through 9) on Page 23 are unworkable. They should be dropped entirely. We
  contacted the authors of the reference document and asked if the software to
  perform these tests as described on page 23 was available, and here is what
  Ron Judkoff replied in an email to us and Mo Madani:  Ron Judkoff: “In August of 1997, NREL published “Home Energy Rating System Building Energy Simulation Test for Florida (Florida-HERS BESTEST)” in two volumes. Volume 1 was a User’s Manual and Volume 2 contained Reference Results. The reason for publishing this separate version of HERS BESTEST was that Phil Fairey was concerned that we did not emphasize conditions and issues important in the Florida climate enough in the original HERS BESTEST. I lost track of what the state of Florida did with this after it was published, but the requirements for the FY10 Florida code are misguided, inappropriate, and probably impossible to comply with. I don’t know what the immediate remedy is here, but I am happy to weigh in on what a more reasonable requirement should look like.”  | 
  
   Delete “Detail Testing Procedures”  The State of Florida position is not to provide a range or additional reference results at this time. This burden remains on the Vendor. The following shall be deleted from the manual Detail Testing
  Procedures 5) Determine weather data for the following ten cities (use City Hall as reference address): Pensacola, Tallahassee, Jacksonville, Gainesville, Orlando, West Palm Beach, Miami, Key West, Naples and Tampa 6) Substitute the values in Table 2-1 7) Run cases in the following Programs identified in the reference document a. Blast 3.0 Level 215 b. DOE 2.1 E-W54 c. Serires/Suncode 5.7 8) Repeat the test cases in the proposed compliance software program 9) Determine the pass/fail ranges as specified in Appendix A  | 
  
   Discussion on Orlando, Las Vegas or Colorado data Availability of reference programs a. Blast 3.0 Level 215 b. DOE 2.1 E-W54 c. Serires/Suncode 5.7 Oct. 17, 2011 WG: for testing use Orlando and
  other cities as applicable (i.e. Las Vegas and Colorado) Use Tier 1 and delete the
  reference to Tier 2 – more applicable to passive solar and not applicable to
  Florida. Delete “detail testing procedures”
  as noted in column 2.   | 
 ||
| 
   29  | 
  
   7.1  | 
  
   [JN comment:  Heading General Requirements And Scope has no text.    | 
  
   Move text depicted below from  Heading “Calculation software tools” to General Requirements Calculation procedures used to comply with this section shall be
  only compliance software tools approved by
  the Florida Building Commission to be capable
  of calculating the annual energy consumption of all building elements that
  differ between the standard reference design and the proposed design and shall include the following capabilities.  1.      
  Computer
  generation of the standard
  reference design using
  only the input for the proposed
  design. The
  calculation procedure shall not allow the user to directly modify the
  building component characteristics of the standard reference design. 2.      
  Building
  operation for a full calendar year (8760 hours). 3.      
  Climate
  data for a full calendar year (8760 hours) and shall reflect approved coincident hourly data for
  temperature, solar radiation, humidity and wind speed for the building
  location. 4.      
  Ten
  or more thermal zones. 5.      
  Thermal
  mass effects. 6.      
  Hourly
  variations in occupancy, illumination, receptacle loads, thermostat settings,
  mechanical ventilation, HVAC equipment availability, service hot water usage
  and any process loads. 7.      
  Part-load
  performance curves for mechanical equipment. 8.      
  Capacity
  and efficiency correction curves for mechanical heating and cooling
  equipment. 9.      
  Printed
  code official inspection checklist listing each of
  the proposed design
  component
  characteristics from Table 506.5.1(1) determined by the analysis to provide
  compliance, along with their respective performance ratings (e.g., R-value, U-factor, SHGC, HSPF, AFUE, SEER, EF,
  etc.).  | 
  
   Will add section number s to remaining headings for
  additional clarity Oct. 17, 2011 WG:  agree - add section numbers to remaining
  headings for additional clarity.  | 
 ||
| 
   35  | 
  
   7.2  | 
  
   ASHRAE Standard 140-2007 Tests JN comment: 140-2011
  should be published by ASHRAE in late October – consider updating to that,
  and if yes, cite “140-2011 Class I tests of Section 5”]  | 
  
   Acceptable  | 
  
   Need copy of 140-2011 Oct. 17, 2011 WG:  Use 140-2007 for consistency with the FBC,
  Energy Conservation.   | 
 ||
| 
   35  | 
  
  7.2
   | 
  
   JN comment: COMNET compliant software is required to perform the ASHRAE Standard 140-2007 suite of software tests and the results of these tests shall conform to the COMNET acceptance requirements [provide the COMNET document reference]. All tests shall be completed in accordance with the requirements of ASHRAE Standard 140-2007. The resulting estimates of energy consumption shall fall between the minimum and maximum values established by COMNET, unless a logical explanation is provided using the standard output report block for “anomalous results” provided with Standard 140-2007 Addendum a [this will also be included in 140-2011]. The portfolio folder for Appendix E contains spreadsheets wherein the software vendor enters the results of the Standard 140 simulations for comparison against the criteria. When results from candidate software fall outside the COMNET acceptance range or when candidate software is unable to perform one of the tests, the vendor shall provide an explanation of the reason as per ASHRAE Standard 140-2007 Addendum A [or -2011] requirements. The portfolio folder for Appendix E also contains a methodology paper that describes how the COMNET acceptance criteria were developed.  | 
  
   COMNET compliant software is
  required to perform the ASHRAE Standard 140-2007 suite of software tests and
  the results of these tests shall conform to the COMNET acceptance
  requirements [provide the COMNET document reference]. All tests shall
  be completed in accordance with the requirements of ASHRAE Standard 140-2007.
  The resulting estimates of energy consumption shall fall between the minimum
  and maximum values established by COMNET, unless a logical explanation
  is provided using the standard output report block for “anomalous results”
  provided with Standard 140-2007 Addendum a
  [this will also be included in 140-2011]. The portfolio folder for
  Appendix E contains spreadsheets wherein the software vendor enters the
  results of the Standard 140 simulations for comparison against the criteria.
  When results from candidate software fall outside the COMNET acceptance range
  or when candidate software is unable to perform one of the tests, the vendor
  shall provide an explanation of the reason as per ASHRAE Standard 140-2007 Addendum
  A [or -2011] requirements. The portfolio folder for Appendix E also
  contains a methodology paper that describes how the COMNET acceptance
  criteria were developed.  | 
  
   Insert COMMET document reference Underlined text added Update reference to depict Standard 140-2007 Addendum a 
 
 Oct. 17, 2011 WG: for testing auto generation of
  the reference design use the criteria stated in Section B - 1.2 of the 2010
  FBC, Energy Conservation.    | 
 ||
| 
   38  | 
  
  Appendix A
   | 
  
   JN comment: Appendix A Compare Annex B22 starting on p. 134 of attached pdf, which is the current ASHRAE pub galley (internal, not for distribution), versus your App A and make changes to appropriate sections of what you have to match the content changes (obviously formats, section #s, eqn #s, etc would remain in your context not ASHRAE's) Ron Judkoff: One comment….I would not refer
  to the change from N-1 to N in the “Acceptance Criteria” as a correction, but
  as a change. Also, I believe that when we made that change we also changed
  the “example” confidence interval criteria so that the target ranges remained
  the same. JN comment: I agree with you on the
  first note -- we changed to "N" per a comment by an industry
  professional, although one could develop a supporting argument to keep the
  "N-1" in the denominator.    And yes on the second note,
  that's what we did. And the changes to the confidence coefficients for
  Florida-HERS BESTEST should be the same as those for HERS BESTEST,
  so that there are no changes to the acceptance ranges of Florida-HERS BESTEST
  Volume 2, but someone on the Florida side should double check that.  | 
  
   Change the
  equations to reflect only "N" in both numerator and denominator.  | 
  
   Update Appendix A to 140-2011 informative annex  changes Oct. 17, 2011 WG:  Take out Appendix A.  Not needed. WG:  Okay as is “N-1”.  | 
 ||
| 
   38  | 
  
   | 
  
   JN comment: Update this for
  forthcoming 140-2007 Addendum B (will also be included in 140-2011) – check
  with Steve Ferguson if 140-2007-B is posted; I know 140-2007-C is not yet
  posted, but B might be.]  | 
  
   Acceptable  | 
  
   Need copy of 140-2011 Oct. 17, 2011 WG:  Use 140 -2007 for consistency with the
  standard edition as referenced in the FBC, Energy Conservation.   | 
 ||
| 
   42  | 
  
   | 
  
   JN comment:  Appendix B-delete  | 
  
   Appendix B is a hot hyperlink.  Please double click to launch.  This reference will remain in the
  manual.  Appendix B currently contain: Model Validation and Testing: The
  Methodological Foundation of ASHRAE Standard 140  Preprint R.
  Judkoff National Renewable Energy Laboratory J. Neymark J. Neymark
  & Associates Presented at the ASHRAE 2006 Annual Meeting Quebec City,
  Canada June 24–29, 2006  | 
  
    
    | 
 ||
| 
   43  | 
  
   | 
  
   JN comment:  References Judkoff and Neymark 2006 [enter in appropriate format using http://www.nrel.gov/docs/fy06osti/40360.pdf Same content as 
  | 
  
   Update references to reflect changes  Add COMMET User Manual, ASHRAE 90.1  | 
  
   Need working url or electronic copies of reference
  documents Oct. 17,2011 WG:  this should only be included in the
  reference document list for informational purposes.  | 
 ||
| 
   | 
  
   | 
  
   | 
  
   | 
  
   | 
 ||
| 
   | 
  
   | 
  
   Comments from William Wright  Forms 405 and 506  All forms required for submittal for permits
  (i.e. Form 405, 405, 506) should be shown in detail in the Manual or the
  Code. . It is unclear what data should be presented in these forms, without a
  provided template. Furthermore, allowing vendors to each have different
  formats will result in considerable frustration on the part of Inspectors and
  users   | 
  
   Add Forms to manual (create appendix C and D).  State currently does not have these forms
  as a given template. Form for EPL Card may be useful also.  | 
  
   Need copies of Forms identified in the code.  The EPL
  Display Card may be useful to add also. 405.4.3 Additional documentation. The code official shall require the following
  documents: 1. An EPL Display Card signed by the builder providing the
  building component characteristics of the proposed design shall be provided to the purchaser of the home at time of title
  transfer. Oct. 17, 2011 WG: add an appendix necessary to
  describe Forms 405 and 506.  Use Energy
  Gauge printout to develop templates for these forms.  Also, include a copy of the EPL Display
  Card in the appendix.     | 
 ||
| 
   | 
  
   | 
  
   Comments from William Wright  Table
  B2.2 from Normative Appendix B in the revised version of the Energy Code is
  referenced by the Manual, and requires knowledge contained in Energy Gauge
  (See section on Building Envelope, Standard Reference Design, Table B2.2). We
  assume these references to Energy Gauge should be removed.   | 
  
   State of FL states that the reference to Energy Guage was
  removed  from the current draft of the
  2010 Florida Building Code, Energy Conservation  | 
  
   Need the latest draft of the Florida building code Oct. 17, 2011 Note:  No action needed.  Staff stated that this has been corrected.  | 
 ||
| 
   | 
  
   23  | 
  
   Comments from William Wright  Table 2-1 without accompanying references  | 
  
   Scrub entire draft to add references where necessary.  | 
  
   Oct. 17, 2011 WG:  Agree with the comment in column 2. Utility rates for the commercial
  option are available for the FPSC.   | 
 ||
| 
   | 
  
   | 
  
   Additional comments submitted during
  the meeting  | 
  
   | 
  
   | 
 ||
| 
   9  | 
  
   1.5  | 
  
   Copy of the Compliance…….A machine
  readable copy……zones.  | 
  
   | 
  
   Oct. 17, 2011 WG: change “machine readable” to
  “pdf”. Note: need to differentiate
  between the term “Energy Code” and “program code”.  | 
 ||
| 
   19  | 
  
   5.1  | 
  
   Paragraph – “The compliance software shall automatically
  ….the standard reference design to establish the energy budget.  | 
  
   | 
  
   Oct. 17, 2011 WG:  delete entire paragraph.  This section is more applicable to equipment
  sizing.  | 
 ||
| 
   20  | 
  
   2nd   | 
  
   2nd Paragraph – clarify “expected results”  | 
  
   | 
  
   Oct. 17, 2011 WG: clarify text by referring to
  “BESTEST”.  |